Hays County
- Asia > Middle East > Iran (0.16)
- Asia > China (0.07)
- Asia > Middle East > Jordan (0.05)
- (3 more...)
- Media > News (1.00)
- Government > Regional Government > North America Government > United States Government (1.00)
- Government > Military > Navy (1.00)
- Information Technology > Artificial Intelligence > Robots > Autonomous Vehicles > Drones (0.47)
- Information Technology > Artificial Intelligence > Robots > Locomotion (0.42)
Learning to Forget with Information Divergence Reweighted Objectives for Noisy Labels
Birrell, Jeremiah, Ebrahimi, Reza
We introduce ANTIDOTE, a new class of objectives for learning under noisy labels which are defined in terms of a relaxation over an information-divergence neighborhood. Using convex duality, we provide a reformulation as an adversarial training method that has similar computational cost to training with standard cross-entropy loss. We show that our approach adaptively reduces the influence of the samples with noisy labels during learning, exhibiting a behavior that is analogous to forgetting those samples. ANTIDOTE is effective in practical environments where label noise is inherent in the training data or where an adversary can alter the training labels. Extensive empirical evaluations on different levels of symmetric, asymmetric, human annotation, and real-world label noise show that ANTIDOTE outperforms leading comparable losses in the field and enjoys a time complexity that is very close to that of the standard cross entropy loss.
- North America > United States > Florida > Hillsborough County > Tampa (0.14)
- Asia > Middle East > Jordan (0.04)
- North America > United States > Texas > Hays County > San Marcos (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
Applying MambaAttention, TabPFN, and TabTransformers to Classify SAE Automation Levels in Crashes
Somvanshi, Shriyank, Tusti, Anannya Ghosh, Mimi, Mahmuda Sultana, Islam, Md Monzurul, Polock, Sazzad Bin Bashar, Dutta, Anandi, Das, Subasish
The increasing presence of automated vehicles (AVs) presents new challenges for crash classification and safety analysis. Accurately identifying the SAE automation level involved in each crash is essential to understanding crash dynamics and system accountability. However, existing approaches often overlook automation-specific factors and lack model sophistication to capture distinctions between different SAE levels. To address this gap, this study evaluates the performance of three advanced tabular deep learning models MambaAttention, TabPFN, and TabTransformer for classifying SAE automation levels using structured crash data from Texas (2024), covering 4,649 cases categorized as Assisted Driving (SAE Level 1), Partial Automation (SAE Level 2), and Advanced Automation (SAE Levels 3-5 combined). Following class balancing using SMOTEENN, the models were trained and evaluated on a unified dataset of 7,300 records. MambaAttention demonstrated the highest overall performance (F1-scores: 88% for SAE 1, 97% for SAE 2, and 99% for SAE 3-5), while TabPFN excelled in zero-shot inference with high robustness for rare crash categories. In contrast, TabTransformer underperformed, particularly in detecting Partial Automation crashes (F1-score: 55%), suggesting challenges in modeling shared human-system control dynamics. These results highlight the capability of deep learning models tailored for tabular data to enhance the accuracy and efficiency of automation-level classification. Integrating such models into crash analysis frameworks can support policy development, AV safety evaluation, and regulatory decisions, especially in distinguishing high-risk conditions for mid- and high-level automation technologies.
- North America > United States > Texas > Hays County > San Marcos (0.05)
- North America > United States > West Virginia (0.04)
- North America > United States > California (0.04)
- (4 more...)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (0.67)
- Transportation > Ground > Road (1.00)
- Government (1.00)
- Transportation > Passenger (0.93)
- (2 more...)
Can Large Language Models Challenge CNNs in Medical Image Analysis?
Ahmed, Shibbir, Sakib, Shahnewaz Karim, Das, Anindya Bijoy
This study presents a multimodal AI framework designed for precisely classifying medical diagnostic images. Utilizing publicly available datasets, the proposed system compares the strengths of convolutional neural networks (CNNs) and different large language models (LLMs). This in-depth comparative analysis highlights key differences in diagnostic performance, execution efficiency, and environmental impacts. Model evaluation was based on accuracy, F1-score, average execution time, average energy consumption, and estimated $CO_2$ emission. The findings indicate that although CNN-based models can outperform various multimodal techniques that incorporate both images and contextual information, applying additional filtering on top of LLMs can lead to substantial performance gains. These findings highlight the transformative potential of multimodal AI systems to enhance the reliability, efficiency, and scalability of medical diagnostics in clinical settings.
- North America > United States > Texas > Hays County > San Marcos (0.04)
- North America > United States > Tennessee > Hamilton County > Chattanooga (0.04)
- North America > United States > Ohio > Summit County > Akron (0.04)
- Asia > Myanmar > Tanintharyi Region > Dawei (0.04)
- Health & Medicine > Therapeutic Area > Oncology (1.00)
- Health & Medicine > Diagnostic Medicine > Imaging (1.00)
A Systematic Evaluation of LLM Strategies for Mental Health Text Analysis: Fine-tuning vs. Prompt Engineering vs. RAG
Kermani, Arshia, Perez-Rosas, Veronica, Metsis, Vangelis
This study presents a systematic comparison of three approaches for the analysis of mental health text using large language models (LLMs): prompt engineering, retrieval augmented generation (RAG), and fine-tuning. Using LLaMA 3, we evaluate these approaches on emotion classification and mental health condition detection tasks across two datasets. Fine-tuning achieves the highest accuracy (91% for emotion classification, 80% for mental health conditions) but requires substantial computational resources and large training sets, while prompt engineering and RAG offer more flexible deployment with moderate performance (40-68% accuracy). Our findings provide practical insights for implementing LLM-based solutions in mental health applications, highlighting the trade-offs between accuracy, computational requirements, and deployment flexibility.
- North America > United States > Texas > Hays County > San Marcos (0.04)
- Europe > Belgium > Brussels-Capital Region > Brussels (0.04)
A Survey on Structured State Space Sequence (S4) Models
Somvanshi, Shriyank, Islam, Md Monzurul, Mimi, Mahmuda Sultana, Polock, Sazzad Bin Bashar, Chhetri, Gaurab, Das, Subasish
Recent advancements in sequence modeling have led to the emergence of Structured State Space Models (SSMs) as an efficient alternative to Recurrent Neural Networks (RNNs) and Transformers, addressing challenges in long-range dependency modeling and computational efficiency. While RNNs suffer from vanishing gradients and sequential inefficiencies, and Transformers face quadratic complexity, SSMs leverage structured recurrence and state-space representations to achieve superior long-sequence processing with linear or near-linear complexity. This survey provides a comprehensive review of SSMs, tracing their evolution from the foundational S4 model to its successors like Mamba, Simplified Structured State Space Sequence Model (S5), and Jamba, highlighting their improvements in computational efficiency, memory optimization, and inference speed. By comparing SSMs with traditional sequence models across domains such as natural language processing (NLP), speech recognition, vision, and time-series forecasting, we demonstrate their advantages in handling long-range dependencies while reducing computational overhead. Despite their potential, challenges remain in areas such as training optimization, hybrid modeling, and interpretability. This survey serves as a structured guide for researchers and practitioners, detailing the advancements, trade-offs, and future directions of SSM-based architectures in AI and deep learning.
- North America > United States > Texas > Hays County > San Marcos (0.04)
- North America > United States > California > Santa Clara County > Santa Clara (0.04)
- Asia > Middle East > Jordan (0.04)
- Europe > Italy > Calabria > Catanzaro Province > Catanzaro (0.04)
Crash Severity Analysis of Child Bicyclists using Arm-Net and MambaNet
Somvanshi, Shriyank, Chakraborty, Rohit, Das, Subasish, Dutta, Anandi K
Child bicyclists (14 years and younger) are among the most vulnerable road users, often experiencing severe injuries or fatalities in crashes. This study analyzed 2,394 child bicyclist crashes in Texas from 2017 to 2022 using two deep tabular learning models (ARM-Net and MambaNet). To address the issue of data imbalance, the SMOTEENN technique was applied, resulting in balanced datasets that facilitated accurate crash severity predictions across three categories: Fatal/Severe (KA), Moderate/Minor (BC), and No Injury (O). The findings revealed that MambaNet outperformed ARM-Net, achieving higher precision, recall, F1-scores, and accuracy, particularly in the KA and O categories. Both models highlighted challenges in distinguishing BC crashes due to overlapping characteristics. These insights underscored the value of advanced tabular deep learning methods and balanced datasets in understanding crash severity. While limitations such as reliance on categorical data exist, future research could explore continuous variables and real-time behavioral data to enhance predictive modeling and crash mitigation strategies.
- North America > United States > Texas > Hays County > San Marcos (0.05)
- Asia > Middle East > Israel (0.05)
- North America > United States > New York (0.04)
- (4 more...)
- Leisure & Entertainment > Sports > Cycling (0.88)
- Transportation (0.69)
Applying Tabular Deep Learning Models to Estimate Crash Injury Types of Young Motorcyclists
Somvanshi, Shriyank, Tusti, Anannya Ghosh, Chakraborty, Rohit, Das, Subasish
Young motorcyclists, particularly those aged 15 to 24 years old, face a heightened risk of severe crashes due to factors such as speeding, traffic violations, and helmet usage. This study aims to identify key factors influencing crash severity by analyzing 10,726 young motorcyclist crashes in Texas from 2017 to 2022. Two advanced tabular deep learning models, ARMNet and MambaNet, were employed, using an advanced resampling technique to address class imbalance. The models were trained to classify crashes into three severity levels, Fatal or Severe, Moderate or Minor, and No Injury. ARMNet achieved an accuracy of 87 percent, outperforming 86 percent of Mambanet, with both models excelling in predicting severe and no injury crashes while facing challenges in moderate crash classification. Key findings highlight the significant influence of demographic, environmental, and behavioral factors on crash outcomes. The study underscores the need for targeted interventions, including stricter helmet enforcement and educational programs customized to young motorcyclists. These insights provide valuable guidance for policymakers in developing evidence-based strategies to enhance motorcyclist safety and reduce crash severity.
- North America > United States > Texas > Hays County > San Marcos (0.04)
- North America > Canada > British Columbia (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- (6 more...)
- Health & Medicine (1.00)
- Transportation > Ground > Road (0.96)
- Government (0.88)
Positional Encoding in Transformer-Based Time Series Models: A Survey
Irani, Habib, Metsis, Vangelis
As a result, machine learning-based approaches, particularly Recent advancements in transformer-based models Recurrent Neural Networks (RNNs) and Convolutional have greatly improved time series analysis, providing Neural Networks (CNNs), have gained popularity robust solutions for tasks such as forecasting, due to their ability to model complex temporal anomaly detection, and classification. A crucial element dynamics [4, 24]. of these models is positional encoding, which RNNs, including their more advanced variants like allows transformers to capture the intrinsic sequential Long Short-Term Memory (LSTM) and Gated Recurrent nature of time series data. This survey systematically Units (GRU), excel at modeling sequential data examines existing techniques for positional encoding by maintaining a hidden state that captures information in transformer-based time series models. We investigate from previous time steps. These architectures a variety of methods, including fixed, learnable, offer several advantages for time series analysis: they relative, and hybrid approaches, and evaluate naturally handle irregular time intervals and missing their effectiveness in different time series classification data points through their sequential processing, tasks. Furthermore, we outline key challenges excel at capturing local temporal patterns through and suggest potential research directions to enhance their recurrent connections, and exhibit a beneficial positional encoding strategies. By delivering a comprehensive "recency bias" where recent time steps are weighted overview and quantitative benchmarking, more heavily than distant ones--a characteristic particularly this survey intends to assist researchers and practitioners valuable in applications like financial forecasting in selecting and designing effective positional and weather prediction. However, RNNs suffer encoding methods for transformer-based time series from inherent limitations such as vanishing and models. The source code for the methods and experiments exploding gradients, making it difficult to learn dependencies discussed in this survey is available on over long time horizons [22].
- North America > Trinidad and Tobago > Trinidad > Arima > Arima (0.04)
- North America > United States > Texas > Hays County > San Marcos (0.04)
An Interactive Framework for Implementing Privacy-Preserving Federated Learning: Experiments on Large Language Models
Ahmadi, Kasra, Behnia, Rouzbeh, Ebrahimi, Reza, Kermani, Mehran Mozaffari, Birrell, Jeremiah, Pacheco, Jason, Yavuz, Attila A
Federated learning (FL) enhances privacy by keeping user data on local devices. However, emerging attacks have demonstrated that the updates shared by users during training can reveal significant information about their data. This has greatly thwart the adoption of FL methods for training robust AI models in sensitive applications. Differential Privacy (DP) is considered the gold standard for safeguarding user data. However, DP guarantees are highly conservative, providing worst-case privacy guarantees. This can result in overestimating privacy needs, which may compromise the model's accuracy. Additionally, interpretations of these privacy guarantees have proven to be challenging in different contexts. This is further exacerbated when other factors, such as the number of training iterations, data distribution, and specific application requirements, can add further complexity to this problem. In this work, we proposed a framework that integrates a human entity as a privacy practitioner to determine an optimal trade-off between the model's privacy and utility. Our framework is the first to address the variable memory requirement of existing DP methods in FL settings, where resource-limited devices (e.g., cell phones) can participate. To support such settings, we adopt a recent DP method with fixed memory usage to ensure scalable private FL. We evaluated our proposed framework by fine-tuning a BERT-based LLM model using the GLUE dataset (a common approach in literature), leveraging the new accountant, and employing diverse data partitioning strategies to mimic real-world conditions. As a result, we achieved stable memory usage, with an average accuracy reduction of 1.33% for $\epsilon = 10$ and 1.9% for $\epsilon = 6$, when compared to the state-of-the-art DP accountant which does not support fixed memory usage.
- North America > United States > Florida > Hillsborough County > Tampa (0.14)
- North America > United States > Florida > Hillsborough County > University (0.04)
- North America > United States > Texas > Hays County > San Marcos (0.04)
- (3 more...)